In many cases, the 'labeled' outcome is difficult to observe and may require a complicated or expensive procedure, and the predictor information is easy to be obtained. We propose a semi-supervised estimator for the one-dimensional varying coefficient regression model which improves the conventional supervised estimator by using the unlabeled data efficiently. The semi-supervised estimator is proposed by introducing the intercept model and its asymptotic properties are proven. The Monte Carlo simulation studies and a real data example are conducted to examine the finite sample performance of the proposed procedure.
Jun Ma, Xiaolong Zhu .
Robust safe semi-supervised learning framework for high-dimensional data classification. AIMS Mathematics, 2024, 9(9): 25705-25731.
doi: 10.3934/math.20241256
[6]
Yanting Xiao, Wanying Dong .
Robust estimation for varying-coefficient partially linear measurement error model with auxiliary instrumental variables. AIMS Mathematics, 2023, 8(8): 18373-18391.
doi: 10.3934/math.2023934
[7]
Qiang Zhao, Zhaodi Wang, Jingjing Wu, Xiuli Wang .
Weighted expectile average estimation based on CBPS with responses missing at random. AIMS Mathematics, 2024, 9(8): 23088-23099.
doi: 10.3934/math.20241122
[8]
Wenhui Feng, Xingfa Zhang, Yanshan Chen, Zefang Song .
Linear regression estimation using intraday high frequency data. AIMS Mathematics, 2023, 8(6): 13123-13133.
doi: 10.3934/math.2023662
[9]
Junfeng Huo, Mingquan Wang, Xiuqing Zhou .
Nonparametric estimation of coefficient function derivatives in varying coefficient models. AIMS Mathematics, 2025, 10(5): 11592-11626.
doi: 10.3934/math.2025526
In many cases, the 'labeled' outcome is difficult to observe and may require a complicated or expensive procedure, and the predictor information is easy to be obtained. We propose a semi-supervised estimator for the one-dimensional varying coefficient regression model which improves the conventional supervised estimator by using the unlabeled data efficiently. The semi-supervised estimator is proposed by introducing the intercept model and its asymptotic properties are proven. The Monte Carlo simulation studies and a real data example are conducted to examine the finite sample performance of the proposed procedure.
1.
Introduction
Semi-supervised learning first appeared in machine learning literature, used to describe a situation where some data are labeled and the rest are unlabeled [10]. Conceptually, it is between supervised and unsupervised learning. It allows you to take advantage of a large amount of unlabeled data available and smaller labeled datasets in many instances. Semi-supervised learning occurs when the label variable is difficult to observe and may require complex or expensive processes. Specifically, a sample of n observations in the joint distribution (X,Y) is given, where Y is the label variable and X contains the covariates. In addition, an additional m samples are observed with only X given. The aim is to study the relationships between X and Y using additional unlabeled data.
For semi-supervised learning, much literature focuses on the case that Y takes a small number of values, which can be reduced to the case of classification tasks, such as [19]. In recent years, research in this field has focused on neural network-based models and generative learning algorithm problems. [13] introduced a simple and computationally efficient algorithm for training deep neural networks in a semi-supervised learning paradigm: interpolating Consistency Training (ICT). [8] introduced an innovative framework for federated transfer learning (FTL) known as Semi-Supervised Federated Heterogeneous Transfer Learning (SFHTL), to utilize unlabeled non-overlapping samples, addressing the challenge of model overfitting caused by the limited overlap of training samples in federated learning (FL) scenarios. Based on compatibility conditions in the semi-supervised probably approximately correct (PAC) theory, [7] demonstrated why labeled heterogeneous source data and unlabeled target data help reduce target risk. Based on this theory, two algorithms were designed as a proof of concept. One is the kernel heterogeneous domain alignment (KHDA) algorithm, which is a kernel-based algorithm, and the other is the joint mean embedded alignment (JMEA) algorithm based on neural networks. At the same time, some scholars have studied the regression problem. [14] classified different semi-supervised methods into two categories: distribution-based and marginal-based. The distribution method relied on the assumption that the conditional expectation E(Y|X) was linked to the marginal distribution of X. The marginal-based approach used additional information on X. Some other studies, such as [18], took into account Y sequential values and used unlabeled data to learn the structure of X so that non-parametric regression could be better estimated. These efforts are very helpful where non-parametric regression is useful and unlabeled data is available. [15] proposed an estimator of the population mean using unlabeled data combined with least squares. [16] provided a semi-supervised reasoning framework focused on the mean and variance of the responses, allowing covariates to be much larger in size than the sample sizes, and provided new estimates of the mean and variance of response variables. [4] considered the linear regression problem in a semi-supervised setting and proposed a class of highly efficient adaptive semi-supervised estimators (EASE) to improve the estimation efficiency. Additionally, they applied this method to the study of electronic medical records on autoimmunity. [2] proposed a correction estimator that effectively integrates labeled and unlabeled data, called Corrected High-dimensional Inference of Variance Interpretation Estimators (CHIVE), which achieved minimax optimal convergence speed under a general semi-supervised framework. [1] studied linear regression in a semi-supervised setting and, even in a framework where E(Y|X) was not linear, additional information about the distribution of X was helpful to construct better estimates than standard least squares estimates. However, there are few studies on semi-supervised learning in the varying coefficient models. Therefore, we want to extend the intercept model method of [1] to the varying coefficient model and investigate the estimation of the coefficient function under semi-supervised learning.
Varying coefficient models constitute a versatile and expansive category of statistical models encompassing well-known structures like additive models, partial linear models, single-index coefficient regression models and adaptive varying coefficient partial linear models. Researchers have investigated varying coefficient models, seeking to understand and refine their applications across diverse statistical contexts. The exploration of these models extends beyond conventional frameworks, uncovering novel insights that contribute to the evolving landscape of statistical methodology. [6] introduced a robust two-step estimation approach designed for the coefficient function. The methodology involves the construction of a reliable estimator for the coefficient function, accompanied by a thorough analysis of the asymptotic mean squared error and the convergence velocity of the estimator. This method not only advances the precision of coefficient function estimation but also contributes insights into the statistical properties of the estimator under consideration. [3] investigated the generalized varying coefficient model, focusing on both estimation and hypothesis testing. Their methodological approach centered on crafting a robust estimation for the coefficient function, leveraging the local polynomial regression technique to enhance accuracy. [9] proposed a semiparametric estimator for the heteroscedastic single-index varying coefficient model. The estimator is proved to attain the semiparametric efficiency bound. [17] considered an estimating equations approach to parameter estimation in an adaptive varying coefficient linear quantile model. They proposed estimating equations for the index vector of the model in which the unknown nonparametric functions were estimated by minimizing the check loss function, resulting in a profiled approach.
Our goal here is to study the more efficient estimates of the one-dimensional varying coefficient model with unlabeled data as a pioneer in the semi-supervised varying coefficient modeling problems. The intercept model is introduced into the varying coefficient models, and the extra information from the unlabeled data is combined to improve the performance of the estimators.
The rest of the article is organized as follows. Section 2 provides the basic setting and the proposed methods. Theoretical properties are also given in Section 2. In Section 3, we conduct Monte Carlo simulation studies to examine the finite sample performance of the proposed procedure. We also use the proposed procedure in an example with real data.
2.
Materials and methods
2.1. Varying coefficient model
In the study of the demand for shared bicycles (Y), many studies focus on the impact of the temperature (X), and how the relationships between them may change with time (T). Thus the varying coefficient model is a good choice. Nevertheless, in real-world scenarios, it is common that the labeled variable Y may not be entirely observed, and only observations of the covariate (X,T) are available. In such cases, the introduction of semi-supervised learning becomes imperative. In semi-supervised learning, the data structure involves a set of n observations sampled from the joint distribution (Y,X,T), and an additional m samples are observed with only (X,T) given. The model leverages both labeled and unlabeled data, allowing us to make more informed and robust estimates by incorporating the partially observed labeled samples along with the unlabeled data. This approach is effective when dealing with practical problems where complete observations of the labeled variable Y may be limited.
To study the estimates of the varying coefficient model with unlabeled data, we introduce our basic method and consider the following model,
Y=g(T)X+ε,
(2.1)
where Y is the label variable, X is the one-dimensional covariant, T is the dependent variable, g(⋅) is the unknown measurable function on R and ε is the random error with E(ε|T,X)=0.
2.2. Locally weighted linear regression estimates
In the supervised situation, where the data from model (2.1) are completely observed, the coefficient function g(t) of model (2.1) can be estimated by the locally weighted linear regression estimation method [5]. For any given point t0, use a linear function in a neighborhood of t0,
g(t)≈g(t0)+g′(t0)(t−t0).
(2.2)
We can minimize the following objective functions by a and b to obtain the estimator of g(t0),
L(a,b)=n∑i=1(Yi−aXi−bTi−t0hXi)2Kh(Ti−t0),
(2.3)
where g(t0)=a, hg′(t0)=b and K(⋅) is some kernel function Kh(⋅)=K(⋅/h)/h with bandwidth h. Let Zi=(Xi,Xi(Ti−t0h)) and β(t0)=(a,b)⊤=(g(t0),hg′(t0))⊤. Thus, ˆβ(t0)=(ˆa,ˆb)⊤=(ˆg(t0),hˆg′(t0))⊤ is
ˆβ(t0)=(D⊤KwD)−1(D⊤KwY),
(2.4)
where D is the n×2 matrix and its ith line elements are (Xi,XiTi−t0h), Y=(Y1,Y2,...,Yn)⊤, Kw=diag(Kh(T1−t0),Kh(T2−t0),...,Kh(Tn−t0)). Thus, the locally weighted linear regression estimate for the coefficient function g(t0) of model (2.1) is ˆgL(t0)=e⊤ˆβ(t0), where e=(1,0)⊤.
2.3. Intercept model with total information
To take full advantage of the information from the unlabeled data, we want to discuss the form of estimation in the semi-supervised case. First, we assume that the total information is completely known.
So, we get the following model through multiplying model (2.1) by XE(X2|T),
˜Y=g(T)˜X01+a(T)˜X02+˜ε=˜g(T)˜X1+˜ε,
(2.5)
where ˜Y=YXE(X2∣T), ˜g(T)=(g(T),a(T)), ˜X1=(˜X01,˜X02)⊤, ˜X01=1, ˜X02=X2E(X2∣T)−1 and ˜ε is the remainder term. Under the total information situation, E(X2|T) is known. The multiplication term makes the expectation of ˜Y to be E(˜Y|T)=g(T). Therefore, by introducing this multiplication term into our estimation process, we can leverage the information from the unlabeled data set more effectively.
After introducing the intercept model, we have the locally weighted regression estimator of β(t0),
ˆ˜β(t0)=(˜D⊤Kw˜D)−1(˜D⊤Kw˜Y),
(2.6)
where the ith line elements of ˜D are (˜X⊤i,˜X⊤i(Ti−t0h)) and ~Xi=(˜X01i,˜X02i)⊤, ˜X01i=1, ˜X02i=X2iE(X2|Ti)−1, i=1,2,…,n, ˆ˜β(t0)=(ˆ˜g(t0)hˆ˜g′(t0)). Thus, the total information estimate of ˜g(t0) is ˆgTI(t0)=˜e⊤ˆ˜β(t0), where ˜e=(1,0,0,0)⊤.
2.4. Intercept model with partial information
In practical problems, the overall information is not completely known, so we can only extract partial information through the observed unlabeled data. We consider the estimation under partial information (PI) i.e., the semi-supervised setting.
For semi-supervised data, consider n independent and identically distributed observations (X1,T1,Y1),(X2,T2,Y2),…,(Xn,Tn,Yn) in a joint distribution G, and an additional set (Xn+1,Tn+1),(Xn+2,Tn+2),…,(Xn+m,Tn+m) of m independent observations from the marginal distribution. Model (2.5) becomes
ˇY=g(T)ˇX01+a(T)ˇX02+ˇε=ˇg(T)ˇX2+ˇε,
(2.7)
where ˇY=YXˇE(X2∣T), ˇg(T)=(g(T),a(T)), ˇX2=(ˇX01,ˇX02)⊤, ˇX02=X2ˇE(X2∣T)−1, ˇX01=1 and ˇε is the remainder term. ˇE(X2|T) is the estimated conditional expectation under T for X2 based on the semi-supervised data, ˇE(X2|T)=∑n+mi=1X2iKh(Ti−T)∑n+mi=1Kh(Ti−T). Therefore, the semi-supervised estimator of β(t) is
ˆˇβ(t0)=(ˇD⊤KwˇD)−1(ˇD⊤KwˇY),
(2.8)
where the ith line elements of ˇD are (ˇX⊤i,ˇX⊤i(Ti−t0h)) and ˇXi=(ˇX01i,ˇX02i)⊤, ˇX01i=1, ˇX02i=X2iˇE(X2|Ti)−1, i=1,2,…,n, ˆˇβ(t0)=(ˆˇg(t0)hˆˇg′(t0)). The semi-supervised estimator of the coefficient function ˇg(t0) of the model (2.7) is ˆgPI(t0)=ˇe⊤ˆˇβ(t0), where ˇe=(1,0,0,0)⊤.
2.5. Theoretical properties
In this subsection, the following conditions are listed to study the theoretical properties of the proposed estimating procedures.
(1) The density function f(t) of T is continuous over the interval [0, 1] and takes values greater than 0.
(2) The varying coefficient function g(t) and E(Y|T=t) have continuous derivatives up to order 2, and are bounded away from zero.
(3) The kernel function Kh(⋅) is a bounded symmetric density function satisfying Lipschitz continuity in the interval (-1, 1) with a window width of 0<h and h=O(n−15).
(4) Γ(t)=E(X2|T=t) is non-singular on the interval [0, 1] and has a continuous second derivative.
(5) E(ε2|T=t,X=x) has bounded partial derivatives up to order 2 and is bounded away from zero.
(6) There exists s>2 that makes E(|X|2s)<∞ and E(|Y|2s)<∞.
Remark 1.Conditions (1)–(4) are general assumptions for varying coefficient models, which can be found in Tang and Zhou [12].
Theorem 1.If conditions (1)–(6) hold, and ε(X−E(X|T)) has a finite second order moment, then,
√nh(ˆgL(t0)−g(t0)−12μ2h2g″(t0))∼N(0,σ2L),
where σ2L=ν0f−1(t0)[Γ−1(t0)]2E(X2ε2|T=t0), μ2=∫t2K(t)dt, ν0=∫K2(t)dt.
Proof. First, we consider the varying coefficient model (2.1), known by the properties of the varying coefficient model,
1nD⊤KwDP→(Γ(t0)f(t0)00μ2Γ(t0)f(t0)),
(2.9)
where D is the n×2 matrix and its line i elements are (Xi,XiTi−t0h), Γ(t0)=E(X2|Ti=t0), f(t) is the density function of T and μ2=∫t2K(t)dt. Since
Define the above as Y=Dβ(t0)+ΔgX+ε, where Y=(Y1,Y2,…,Yn)⊤, Δg=diag((g(T1)−g(t0)−g′(t0)(T1−t0)),…,(g(Tn)−g(t0)−g′(t0)(Tn−t0))) and X=(X1,X2,…,Xn)⊤. We have
Theorem 2.Assuming that (˜Y,˜X01,˜X02) has a finite second-order moment, the total information estimator of the intercept model has
√nh(ˆgTI(t0)−g(t0)−12μ2h2g″(t0))∼N(0,σ2TI),
where σ2TI=ν0f−1(t0)E(˜ε2|t0), σ2L=σ2TI+σ2diff1, σ2diff1=nhE{1f(t0)1n∑ni=1Kh(Ti−t0)(εiXiE(X2|Ti)−˜εi)}2.
Proof. Before discussing the form of estimation in the semi-supervised case, we assume that all information is completely known and consider the estimation with complete data information. Consider the intercept model (2.5) under the complete data. We have the objective function
Let the line i elements of ˜D be (˜X⊤i,˜X⊤i(Ti−t0h)), ˜Y=(~Y1,~Y2,⋯,~Yn)⊤, ˆ˜β=(ˆ˜g(t0)hˆ˜g′(t0)), so locally weighted linear regression by varying coefficients shows that ˆ˜β(t0)=(˜D⊤Kw˜D)−1(˜D⊤Kw˜Y), ˆ˜g(t0)=e⊤1ˆ˜β(t0), where the first element of e1 is 1 and the others are 0.
Theorem 3.Considering the partial information estimator of the intercept model in a semi-supervised setting, when limnn+m=v, we have
√nh(ˆgPI(t0)−g(t0)−12μ2h2g″(t0))∼N(0,σ2PI),
where σ2PI=ν0f−1(t0)E(ˇε2|t0), σ2PI=σ2TI+vσ2diff1.
Proof. Considering that in a practical problem the overall information is not completely known, so we can only extract partial information from the observed unlabeled data, we next consider the estimation under partial information (PI). We study the intercept model (2.7) under the semi-supervised setting.
In the same way as (ii), the local linear regression by varying coefficients shows that ˆˇβ(t0)=(ˇD⊤KwˇD)−1(ˇD⊤KwˇY), ˆˇg(t0)=e⊤1ˆˇβ(t0). Where the line i elements of ˇD are (ˇX⊤i,ˇX⊤i(Ti−t0h)) and ˇXi=(ˇX01i,ˇX02i)⊤.
First, let us say
ˇE(X2|T)=∑n+mi=1X2iKh(Ti−T)∑n+mi=1Kh(Ti−T)≜ˇm(T),
and let n+m=N. When T=t, we say m(t)=E(X2|t).
From the properties of nonparametric kernel estimation, we can get
Let ˇX=(ˇX⊤1⋮ˇX⊤n), and considering 1nˇD⊤KwˇXΔga and 1nˇD⊤Kwˇε, let 1nˇD⊤KwˇXΔga=ˇA(1)1, 1nˇD⊤Kwˇε=ˇA(2)2, ˇA1=e⊤1ˇA(1)1, ˇA2=e⊤1ˇA(2)2, where the first element of e1 is 1 and the others are 0. We have
Suppose limnn+m=v, we can get σ2diff2=vσ2diff1, then we have σ2PI=σ2TI+vσ2diff1. □
Remark 2.Theorems 1–3 show that if σ2diff1 is not equal to zero, then σ2L is larger than σ2TI, and, further, if v<1, then σ2L is also larger than σ2PI. This means that the proposed semi-supervised estimators could be more efficient than the supervised estimators.
3.
Simulation studies and results
To study the finite sample performance of the proposed semi-supervised estimates, this section uses some numerical simulations and a real data analysis to compare the estimated performances of the locally weighted linear estimation of the varying coefficient model (LWLR), the total information estimator based on the intercept model (TI) and the semi-supervised estimator under the partial information setting (PI).
where X∼N(0,1), T comes from the uniform distribution U(0,1) and random error δ∼N(0,1). At the same time, ε1 and ε2 are affected by X and T. The labeled samples (Yi,Xi,Ti), i=1,2,...,n and the unlabeled samples (Xi,Ti), i=n+1,n+2,...,n+m are from Model (Ⅰ) or (Ⅱ) with sample size n and m, respectively. Let the kernel function be the Epanechnikov kernel function, K(u)=0.75(1−u2)I(|u|≤1). We determine the bandwidth h of the kernel function through the process of cross-validation (CV). By employing CV, we aim to identify an optimal bandwidth value that enhances the performance of the kernel function in the given context. For numerical simulation, the labeled sample size is n = 10,20,40,60 and the unlabeled sample size is m=n,3n,10n for each model. To compare the pros and cons of the estimators obtained by different methods, the estimation curves are obtained by 1000 repeated simulations, and the estimation effect is evaluated by the root mean squared error (RMSE) and standard deviation (STD) of the estimations, where RMSE=√1q∑qj=11n∑nk=1[ˆgj(tk)−g(tk)]2, and STD=√1n∑nk=11q∑qj=1[ˆgj(tk)−1q∑qj=1ˆgj(tk)]2. The expression "ˆgj(tk)" represents the jth estimation of the coefficient function at the discrete time point tk, where {tk,k=1,2,…,n} refers to the appropriate grid point. The simulation results are shown in Table 1.
Table 1.
RMSE and STD derived in Model (Ⅰ) and (Ⅱ).
As can be seen from Table 1, the proposed total information estimator and the semi-supervised partial information estimator based on the intercept model are generally superior to the ordinary locally weighted linear regression estimator, and the RMSE and STD obtained by the former two methods are generally smaller than the latter one. The proposed semi-supervised intercept model can improve the efficiency of the estimator.
3.2. Real data application
In recent years, the introduction of shared bicycles has enriched the travel and transportation options of urban residents and solved the "last kilometer" travel problem of residents. Shared cycling has a significant impact on building a larger riding community, minimizing greenhouse gas emissions and improving public health and transport issues. But, shared bicycle travel is easily affected by the weather environment, where the temperature is an important disturbance factor. We study the data set of the number of shared bicycles rented hourly in Seoul from 2017.12–2018.11 [11]. There are a total of N = 8760 samples in the data set. To study the superiority of the estimation method proposed in this paper, the variables selected include count of bikes rented at each hour (vehicles, Y), hour of the day (0:00−23:00, T) and temperature (∘C, X). At the same time, n = 1000 labeled samples and m = 7760 unlabeled samples are randomly selected in the original data set to construct a semi-supervised setting, and the semi-supervised intercept model method proposed in this paper is used for analysis. We establish the following model:
Y=β(T)X+ε.
(3.1)
Compare the estimated R2 obtained by semi-supervised intercept model estimation R2PI, the local linear regression estimation R2LWLRN under all supervised data and R2LWLRn under n supervised samples. At the same time, in order to eliminate the influence of randomness, we repeated K=50 times and calculated the mean of the corresponding R2(k) respectively, that is, R2=1K∑Kk=1R2(k). The goodness of fit based on N samples is R2LWLRN=0.8531, the goodness of fit based on n samples is R2LWLRn = 0.7651 and R2PI=0.8411, the effect is significantly better than that of the local linear regression estimation without using unlabeled data.
4.
Conclusions
Semi-supervised data is becoming more common, and most semi-supervised learning methods focus on classification tasks, or solving linear regression models, with less emphasis on varying coefficient models. Therefore, a good estimate of the coefficient function of the varying coefficient model is given in this work. The key idea is to introduce an intercept model to replace the original varying coefficient regression model and perform the estimation under a semi-supervised setting; that is, the information of unlabeled data is utilized in the estimation process. It is further proved that the new estimates have good asymptotic properties. At the same time, the asymptotic properties of the new estimate is better than that of the conventional locally weighted linear regression estimators.
Finally, the method is applied to study the effect of temperature on the demand for shared bicycle rental. The coefficient function of the shared bicycle rental demand model is well estimated, and the demand is predicted.
Use of AI tools declaration
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
Acknowledgments
Peng Lai's research is supported by National Natural Science Foundation of China (11771215). Yanqiu Zhou's research is supported by Guangxi Science and Technology Base and Talent Project (2020ACI9151), and Guangxi University Young and Middle-aged Teachers' Basic Research Ability Improvement Project (2021KY0343).
Conflict of interest
The authors declare that they have no conflicts of interest.
References
[1]
D. Azriel, L. Brown, M. Sklar, R. Berk, A. Buja, L. Zhao, Semi-supervised linear regression, J. Am. Stat. Assoc., 117 (2022), 2238–2251. https://doi.org/10.1080/01621459.2021.1915320 doi: 10.1080/01621459.2021.1915320
[2]
T. Cai, Z. Guo, Semi-supervised inference for explained variance in high dimensional linear regression and its applications, J. R. Stat. Soc. B, 82 (2020), 391–419. https://doi.org/10.1111/rssb.12357 doi: 10.1111/rssb.12357
[3]
Z. Cai, J. Fan, R. Li, Efficient estimation and inferences for varying-coefficient models, J. Am. Stat. Assoc., 95 (2000), 888–902. https://dx.doi.or/10.1080/01621459.2000.10474280 doi: 10.1080/01621459.2000.10474280
[4]
A. Chakrabortty, T. Cai, Efficient and adaptive linear regression in semi-supervised settings, Ann. Statist., 46 (2018), 1541–1572. https://doi.org/10.1214/17-AOS1594 doi: 10.1214/17-AOS1594
J. Fan, W. Zhang, Statistical estimation in varying coefficient models, Ann. Statist., 27 (1999), 1491–1518. https://dx.doi.or/10.1214/aos/1017939139 doi: 10.1214/aos/1017939139
[7]
Z. Fang, L. Lu, F. Liu, G. Zhang, Semi-supervised heterogeneous domain adaptation: theory and algorithms, IEEE T. Pattern Anal., 45 (2023), 1087–1105. https://dx.doi.or/10.1109/TPAMI.2022.3146234 doi: 10.1109/TPAMI.2022.3146234
[8]
S. Feng, B. Li, H. Yu, Y. Liu, Q. Yang, Semi-supervised federated heterogeneous transfer learning, Knowl.-Based Syst., 252 (2022), 109384. https://dx.doi.or/10.1016/j.knosys.2022.109384 doi: 10.1016/j.knosys.2022.109384
[9]
P. Lai, Q. Zhang, H. Lian, Q. Wang, Efficient estimation for the heteroscedastic single-index varying coefficient models, Stat. Probabil. Lett., 110 (2016), 84–93. https://dx.doi.or/10.1016/j.spl.2015.12.005 doi: 10.1016/j.spl.2015.12.005
[10]
C. Merz, D. Clair, W. Bond, SeMi-supervised adaptive resonance theory (SMART2), Proceedings of International Joint Conference on Neural Networks, 1992,851–856. https://dx.doi.org/10.1109/IJCNN.1992.227046 doi: 10.1109/IJCNN.1992.227046
[11]
V. Sathishkumar, J. Park, Y. Cho, Using data mining techniques for bike sharing demand prediction in metropolitan city, Comput. Commun., 153 (2020), 353–366. https://doi.org/10.1016/j.comcom.2020.02.007 doi: 10.1016/j.comcom.2020.02.007
[12]
L. Tang, Z. Zhou, Weighted local linear CQR for varying-coefficient models with missing covariates, TEST, 24 (2015), 583–604. https://doi.org/10.1007/s11749-014-0425-z doi: 10.1007/s11749-014-0425-z
[13]
V. Verma, K. Kawaguchi, A. Lamb, J. Kannala, A. Solin, Y. Bengio, et al., Interpolation consistency training for semi-supervised learning, Neural Networks, 145 (2022), 90–106. https://dx.doi.or/10.1016/j.neunet.2021.10.008 doi: 10.1016/j.neunet.2021.10.008
[14]
J. Wang, X. Shen, W. Pan, On efficient large margin semisupervised learning: method and theory, J. Mach. Learn. Res., 10 (2009), 719–742.
[15]
A. Zhang, L. Brown, T. Cai, Semi-supervised inference: general theory and estimation of means, Ann. Statist., 47 (2019), 2538–2566. https://doi.org/10.1214/18-AOS1756 doi: 10.1214/18-AOS1756
[16]
Y. Zhang, J. Bradic, High-dimensional semi-supervised learning: in search of optimal inference of the mean, Biometrika, 109 (2022), 387–403. https://doi.org/10.1093/biomet/asab042 doi: 10.1093/biomet/asab042
[17]
W. Zhao, J. Li, H. Lian, Adaptive varying-coefficient linear quantile model: a profiled estimating equations approach, Ann. Inst. Stat. Math., 70 (2018), 553–582. https://dx.doi.or/10.1007/s10463-017-0599-8 doi: 10.1007/s10463-017-0599-8
[18]
Z. Zhou, M. Li, Semi-supervised regression with co-training, Proceedings of the 19th international joint conference on Artificial intelligence, 2005,908–913.
[19]
X. Zhu, Semi-supervised learning literature survey, Madison: University of Wisconsin-Madison Publisher, 2005.
This article has been cited by:
1.
Peng Lai, Zhou Wang, Yurong Zhang,
Semi-supervised estimation of a single-index varying-coefficient model,
2025,
218,
01677152,
110312,
10.1016/j.spl.2024.110312