Research article Special Issues

Young returnees' sustainability of return: The case of Bulgaria

  • Received: 14 December 2022 Revised: 23 May 2023 Accepted: 07 June 2023 Published: 14 June 2023
  • Return migration flow to Bulgaria is constant but not high, except for the past two years. Here, we argue that subjective and objective reasons should both be considered in line to explain life prospects of young returnees at home. An innovative approach to balancing the subjective-objective dichotomy in return migration phenomena is measuring sustainability of return through re-emigration, captured by and tested through a re-emigration index based on a national representative survey of returning Bulgarian migrants. The main hypothesis is that the sustainability of the return of young people depends on a complex group of factors, including social and emotional motives. The main findings are about the magnitude of the return migration flow and its sustainability: The factors for the return of young Bulgarians in the pre-Covid research show strong non-economic character, which intensifies in times of crisis. There are no significantly different tendencies for re-emigration/return among young people compared with all returnees.

    Citation: Mihaela Misheva, Antoaneta Getova. Young returnees' sustainability of return: The case of Bulgaria[J]. AIMS Geosciences, 2023, 9(3): 426-440. doi: 10.3934/geosci.2023023

    Related Papers:

    [1] Vanya Ivanova . COVID-19 and return intentions among mobile Bulgarian citizens living abroad. AIMS Geosciences, 2023, 9(2): 382-391. doi: 10.3934/geosci.2023020
    [2] Diana I Bakalova, Ekaterina E Dimitrova . Optimistic expectations and life satisfaction as antecedents of emigration attitudes among Bulgarian Millennials and Zoomers. AIMS Geosciences, 2023, 9(2): 285-310. doi: 10.3934/geosci.2023016
    [3] Jama Mohamed, Dahir Abdi Ali, Abdimalik Ali Warsame, Mukhtar Jibril Abdi, Eid Ibrahim Daud, Mohamed Mohamoud Abdilleh . Bayesian extreme value modelling of annual maximum monthly rainfall in Somalia from 1901 to 2022. AIMS Geosciences, 2024, 10(3): 598-622. doi: 10.3934/geosci.2024031
    [4] Elisabetta Genovese . University student perception of sustainability and environmental issues. AIMS Geosciences, 2022, 8(4): 645-657. doi: 10.3934/geosci.2022035
    [5] Leonardo Mercatanti, Gaetano Sabato . Digital education, geography and multidisciplinarity: Themes, methods and critical issues1. AIMS Geosciences, 2023, 9(1): 184-190. doi: 10.3934/geosci.2023010
    [6] Corey A. Palmer, Katherine P. Markstein, Lawrence H. Tanner . Experimental test of temperature and moisture controls on the rate of microbial decomposition of soil organic matter: preliminary results. AIMS Geosciences, 2019, 5(4): 886-898. doi: 10.3934/geosci.2019.4.886
    [7] Eric Ariel L. Salas, Geoffrey M. Henebry . Canopy Height Estimation by Characterizing Waveform LiDAR Geometry Based on Shape-Distance Metric. AIMS Geosciences, 2016, 2(4): 366-390. doi: 10.3934/geosci.2016.4.366
    [8] Margherita Bufalini, Farabollini Piero, Fuffa Emy, Materazzi Marco, Pambianchi Gilberto, Tromboni Michele . The significance of recent and short pluviometric time series for the assessment of flood hazard in the context of climate change: examples from some sample basins of the Adriatic Central Italy. AIMS Geosciences, 2019, 5(3): 568-590. doi: 10.3934/geosci.2019.3.568
    [9] Shahid Latif, Firuza Mustafa . Trivariate distribution modelling of flood characteristics using copula function—A case study for Kelantan River basin in Malaysia. AIMS Geosciences, 2020, 6(1): 92-130. doi: 10.3934/geosci.2020007
    [10] Shahid Latif, Firuza Mustafa . A nonparametric copula distribution framework for bivariate joint distribution analysis of flood characteristics for the Kelantan River basin in Malaysia. AIMS Geosciences, 2020, 6(2): 171-198. doi: 10.3934/geosci.2020012
  • Return migration flow to Bulgaria is constant but not high, except for the past two years. Here, we argue that subjective and objective reasons should both be considered in line to explain life prospects of young returnees at home. An innovative approach to balancing the subjective-objective dichotomy in return migration phenomena is measuring sustainability of return through re-emigration, captured by and tested through a re-emigration index based on a national representative survey of returning Bulgarian migrants. The main hypothesis is that the sustainability of the return of young people depends on a complex group of factors, including social and emotional motives. The main findings are about the magnitude of the return migration flow and its sustainability: The factors for the return of young Bulgarians in the pre-Covid research show strong non-economic character, which intensifies in times of crisis. There are no significantly different tendencies for re-emigration/return among young people compared with all returnees.



    With the rapid development of big data and artificial intelligence, data-driven predictions now permeate nearly every aspect of modern science [1,2]. Regression analysis, a key component of data-driven science, has wide applications across various domains, including new material design, stock prediction, medical diagnosis, and geological exploration, etc. [3,4,5,6,7,8,9]. Besides, machine learning offers numerous methods for regression analysis, such as support vector regression (SVR), decision trees (DTs), multi-layer perceptrons (MLPs), deep neural networks, etc. Although those methods can achieve high accuracy, the "black box" nature of neural networks limits their interpretability.

    According to the Weierstrass approximation theorem, continuous functions defined on closed intervals can be uniformly approximated by polynomial functions [10]. Moreover, a polynomial function has a simpler form compared to other complex functions, making polynomial regression a natural choice for practical applications [11,12]. Polynomial regression can be divided into univariate function fitting and multi-variate function regression. While univariate function fitting is highly effective supported by popular data analysis software such as Origin, MATLAB, and SPSS based on the least squares algorithm, there remains a lack of fast, effective, and stable algorithms for multivariate function regression.

    The theme of this paper is to design a fast, efficient and accurate algorithm to solve the multivariate polynomial regression problems (Eq (1.1)).

    y=f(x1,x2,,xn)=ni=1αixβii+nj,k=1,jkγjkxμjjxνkk+ns,p,q=1,spqηspqxσssxλppxτqq+ (1.1)

    In the regression process, SVR introduces a kernel function to construct the nonlinear relation. The MLP method employs the weighted summation of variables (iωjixi+θj) to achieve the aggregation of information. As the weighted summation only passes a linear relationship between variables, the nonlinear knowledge, for example the product relation xμjjxνkk, should be approximated by using an activation function, such as the sigmoid function, f(x)=1/(1+ex), the tanh function, f(x)=(exex)/(ex+ex), etc. However, the applications of kernel functions in SVR and activation functions in MLPs make it difficult to interpret the machine learning model.

    Recently, Kolmogorov-Arnold networks (KANs) have gained attention due to their superior accuracy and interpretability compared to multi-layer perceptrons [13]. MLPs have fixed activation functions on nodes while KANs have learnable activation functions on edges. The Kolmogorov-Arnold representation theorem established that, if f is a multivariate continuous function on a bounded domain, then f can be written as a finite composition of continuous functions of a single variable and the binary operation of addition. The regression model of KANs is in the form of

    y=f(x1,x2,,xn)=nL1iL1=1ϕL1,iL,iL1(nL2iL2=1ϕL2,iL1,iL2(n2i2=1ϕ2,i3,i2(n1i1=1ϕ1,i2,i1)(ni0=1ϕ0,i1,i0(xi0)))) (1.2)

    where, L represents the number of network layers and ϕk,ik+1,ik is a continuous functions of signal variable. That is to say, KANs train and select suitable signal variable continuous functions in each network layer from the given basic function library including polynomial functions, exponential functions, logarithmic functions, etc. Although, KANs have higher accuracy and interpretability, the regression of product relation in multivariate polynomial functions still need to transform to the coupling of univariate functions, which will undoubtedly increase the complexity of the regression.

    The traditional artificial neural network uses nodes whose information transmission is via a sigma unit, i.e., a linear weighted sum of inputs, a=iωixi. A pi unit is constructed by replacing the weighted summation in the activation with a weighted product, a=ixωii. Units of this type are also designated "higher order" in that they contain polynomial terms of order greater than one (linear). The simplest way of modeling such a product relation in multivariate polynomial functions is to increase the complexity of the node by a pi unit. Moreover, a sigma-pi unit can formulate a sum of product terms,

    y=mj=1cjaj=mj=1cjni=1xωjii. (1.3)

    Inspired by this architecture of a sigma-pi unit, we find the high-order sigma-pi neural network (SPNN) [14,15] can efficiently achieve the regression of multivariate polynomial functions. The coefficients and exponents of the multivariate polynomial can be determined through the trained neural network weight parameters (cj and ωji).

    Performing the regression of multivariate polynomial functions by SPNN has the following advantages compared with that of MLPs and KANs.

    ● Polynomial regression by the SPNN model has better interpretability than that of MLPs, due to the fact that high-order terms are generated by pi units instead of complex activation functions.

    ● Polynomial regression by the SPNN model has faster learning efficiency compared with that of KANs, owing to the application of a fixed activation function in polynomial form.

    ● The optimization and control of the parameters in the SPNN model are more convenient than that in MLPs and KANs. Besides, the forms of the SPNN model are more friendly to real applications.

    In an MLP neural network, information from different input signals is transmitted through weighted summation. For example, in a two-layer forward neural network (see Figure 1(a)), the hidden layer

    hi=σ1(nj=1αijxj+bi), (2.1)
    Figure 1.  Different structure units in neural networks: (a) sigma-sigma units in the MLP network; (b) sigma-pi units in the high-order network.

    and the output signal

    y=σ2(mi=1βihi+θ), (2.2)

    where σ1 is the activation function of the hidden layer, σ2 is the activation function of the output layer, α is the weight of the first layer network, β is the weight of the second layer network, and b and θ are the biases. The information transmission from xj to hi, and from hi to y, are all by weighted summation. So, this kind of structure unit in MLP neural networks can be called sigma-sigma units.

    To embed high-order coupled information into a neural network, high-order neural networks are designed. For instance, the sigma-pi neural network (SPNN) enables the embedding of product information into the network. In SPNN, the hidden layer

    gi=σ1(nj=1xαijj+bi), (2.3)

    and the output layer

    y=σ2(mi=1βigi+θ). (2.4)

    The information transmission from xj to gi is by weighted product, and from gi to y is by weighted summation, see Figure 1(b).

    As previously mentioned, the basic unit of SPNN is the power function of input signals. We aim to achieve multivariate polynomial regression using the explainable SPNN. In this model, if the activation functions, σ1 and σ2, are linear, the output signal

    y=mi=1βigi+θ=mi=1βi(nj=1xαijj+bi)+θ=mi=1βi(nj=1xαijj)+mi=1βibi+θ. (2.5)

    Cleary, Eq (2.5) is a multivariate polynomial function, where βi is the polynomial coefficient, αij is the polynomial exponent, and mi=1βibi+θ is the constant term. Thus, this explainable SPNN is essentially a type of Kolmogorov-Arnold network.

    To illustrate the regression process of a multivariate polynomial function more clearly, we present a simple regression problem as an example:

    y=f(x1,x2,,xn)= a1xb11+a2xb22++anxbnn+ni,j=1,ijcijxpiixqjj. (2.6)

    Here, {x1,x2,,xn} represents the state variables, and y is the target variable. Since Eq (2.6) lacks a constant term, bias parameters are not used in the designed network. A two-layer SPNN is employed to solve this regression problem. The number of neurons in the input layer equals n, while the hidden layer contains n+n(n1)=n2 neurons. The neural connections from the input layer to the hidden layer primarily capture the regression of power functions, such as xbii and xpiixqjj, corresponding to each term in the polynomial function. The network then performs a weighted summation of all terms from the hidden layer to the output layer.

    Additionally, to accelerate the parameter optimization, we convert the above sigma-pi neural network to the sigma-sigma neural network, and use the back-propagation algorithm to update the parameters. The input signals are first transformed using logarithmic functions,

    xiui={log(xi),xi>0;,xi=0;log(xi),xi<0. (2.7)

    Next, a sigma unit operates on the ui,

    vj=ni=1αjiui=ni=1log(|xi|αji). (2.8)

    Here αji represents the weight from the ith neuron in the input layer to the jth neuron in the hidden layer. A signed exponential function is then applied to vj,

    hj=sig(xi)exp(vj)=sig(xi)exp(ni=1log(|xi|αji))=ni=1xαjii. (2.9)

    Therefore, the conversion from the pi unit to the sigma unit can be achieved using logarithmic and exponential operators. The output signal, y, is then obtained by the sigma unit from the hidden layer to the output layer.

    y=mj=1βjhj=mj=1βjni=1xαjii=mj=1βjsig(xi)exp(ni=1αjilog(|xi|)). (2.10)

    Here, βj is the weight from the jth neuron in the hidden layer to the output signal. The model error is defined as

    Err=12(yˆy)2, (2.11)

    where y is the model output and ˆy is the real value. For parameter updates, the gradient descent method is employed,

    βj=βjηΔβj,αji=αjiηΔαji. (2.12)

    Here, η is the learn rate,

    Δβj= Errβj= Erryyβj= (yˆy)hj, (2.13)

    and

    Δαji= Errαji=Errvjvjαji= Erryyhjhjvjvjαji= (yˆy)βjexp(vj)sig(xi)log(|xi|). (2.14)

    The optimized parameters β and α are the corresponding multivariate polynomial regression coefficients and exponents.

    The pseudo code of multivariate polynomial regression by our designed neural network is presented in Algorithm 1 below.

    Algorithm 1 Multivariate polynomial regression by the explainable sigma-pi neural network
    Inputs: the corresponding multi-variables, x1,x2,,xn, the target variable, y, and the desired multi-variable polynomial form, for example, y=a1xb11+a2xb22++anxbnn+ni,j=1,ijcijxpiixqjj.
    Outputs: the regression coefficients and exponents, such as, ai,bi,cij,pi,qj.
    1: Divide the data into the train set, valid set, and test set.
    2: Construct the sigma-pi neural network based on the desired empirical formula.
    3: Variable conversion, uixi.
    4: Let the sigma unit work on ui, s.t., uivj.
    5: The signed exponentiation effect on vj, s.t., vjhj.
    6: Let the sigma unit work on hj, s.t., hjy, compute the regression error, Err=12(yˆy)2, set the error threshold as ε, and define the total epochs as N.
    7: while epoch<N do
    8:   Parameters update, βj=βjηΔβj,αji=αjiηΔαji.
    9:   if Err<ε then
    10:     break;
    11:     output the weight parameters, αjibi,pi,qj and βjai,cij.
    12:   else if Err>ε then
    13:     continue;
    14:   end if
    15: end while

    In this section, we present two examples to demonstrate the feasibility and effectiveness of the proposed model. Moreover, we illustrate the application of our method in analyzing multivariate correlated regression, specifically concerning the maximum stress on concrete pipes under traffic loads.

    y=2x21x2+3x31x2+4x21x22. (3.1)

    In this example, the model is used to determine the regression coefficients and exponents of Eq (3.1) based on generated data. The variable x1 is linearly sampled from the interval [1,3], x2 is randomly sampled from the interval [1,4], and y is calculated according to Eq (3.1). These generated data are briefly presented in Figure 2, where Figure 2(a) gives the time series of x1, x2, and y; Figure 2(b), (c) display scatter plots of y versus x1, and y versus x2, respectively.

    Figure 2.  Generated data for integer order polynomial regression: (a) the time series of x1, x2, and y; (b) the scatter plot of y versus x1; (c) the scatter plot of y versus x2.

    The number of neurons in the hidden layer equals 3, i.e., the number of terms of polynomial function Eq (3.1). Let u=log(x1),v=log(x2), and the signals u and v are fed into the sigma unit as inputs. It should be noted that the variables x1 and x2 in this example are restricted to positive values. We use the linear activation function in each architecture, thus, the neurons in the hidden layer satisfy

    h1= α11u+α12v=log(xα111)+log(xα122),h2= α21u+α22v=log(xα211)+log(xα222),h3= α31u+α32v=log(xα311)+log(xα322).

    So, by the exponential operation,

    eh1= elog(xα111)+log(xα122)=xα111xα122,eh2= elog(xα211)+log(xα222)=xα211xα222,eh3= elog(xα311)+log(xα322)=xα311xα322.

    The output signal is the weighted sum of eh1,eh2, and eh3 according to the sigma unit from the hidden layer to the output layer.

    y=β1eh1+β2eh2+β3eh3=β1xα111xα122+β2xα211xα222+β3xα311xα322.

    Here, βi represents the corresponding polynomial regression coefficient, and αij denotes the related polynomial regression exponent. Since we are addressing an integer-order polynomial regression problem, the parameters αij are constrained to integer values during the parameter iteration process. The training is conducted for 500 epochs, with a maximum learning rate of 0.01. Compared with other optimizers, such as BGD, SGD, and MBGD, the combination of the Adam optimizer with a one-cycle scheduler has faster convergence speed and better modeling performance, so we chose this optimization algorithm in the training process. The regression performance is evaluated using the mean squared error (MSE) and the coefficient of determination (R2),

    MSE=1nni=1(yi^yi)2,R2=1ni=1(yi^yi)2ni=1(yiˉy)2.

    Here, yi is the real value, ^yi is the regression value, and ˉy is the mean value of y. The modeling results are shown in Figure 3, the train loss and valid loss are depicted by the MSE, and the coefficient of determination R2=1.0.

    Figure 3.  The modeling results: (a) the train loss and valid loss depicted by MSE; (b) the plot of real values versus regression values, the coefficient of determination R2=1.0.

    The weight parameters of the proposed neural network are listed in Table 1. The parameter αij is related to the polynomial exponent, and parameter βi corresponds to the polynomial coefficient.

    Table 1.  The values of weight parameters of our designed SPNN.
    Parameters α11 α12 α21 α22 α31 α32 β1 β2 β3
    Values 2 1 3 1 2 2 2.0 3.0 4.0

     | Show Table
    DownLoad: CSV

    From the results above, it is evident that our designed SPNN effectively solves the integer-order multivariate polynomial regression problem. The coefficients and exponents of the integer-order multivariate polynomial function align with the weight parameters of the designed SPNN. In the next example, we use the SPNN to solve a fractional-order multivariate polynomial regression problem.

    y=2x2.21x1.52+3x3.41x1.02+4x2.81x2.02. (3.2)

    In this example, we use the designed SPNN to solve the fractional-order polynomial regression problem. The values of x1 and x2 are generated in the same manner as in Example 1, and the output signal is computed according to Eq (3.3). The dataset contains 3000 samples, and the training is conducted for 10,000 epochs. The Adam optimizer is employed combined with a one-cycle scheduler. The maximum learning rate is set to 0.009, while the optimizer's learning rate is initialized at 0.004. The corresponding time series are plotted in Figure 4(a), and the regression result is shown in Figure 4(b).

    Figure 4.  The fractional-order polynomial regression: (a) the generated data of the fractional-order polynomial function; (b) the plot of real values versus regression values.

    The values of mean squared error (MSE), coefficient of determination (R2), and corresponding weight parameters are listed in Table 2.

    Table 2.  The values of MSE, R2, and weight parameters.
    Parameters MSE R2 α11 α12 α21 α22 α31 α32 β1 β2 β3
    Values 6.2×105 1.0 2.21 1.54 2.81 2.01 3.39 1.00 2.05 3.87 3.07
    Approximation 2.2 1.5 2.8 2.0 3.4 1.0 2 4 3

     | Show Table
    DownLoad: CSV

    From the regression results, we observe small discrepancies between the actual polynomial coefficients (or exponents) and the weight parameters. However, the parameters closely align with the polynomial coefficients and exponents when considering their approximate values. Thus, we conclude that the designed SPNN can effectively solve fractional-order multivariate polynomial regression problems.

    In this part, we apply the designed SPNN to analyze the maximum stress of concrete sewage pipelines under the combined influence of traffic load, earth pressure, and groundwater level, with a consideration of 12 physical parameter variables. These variables encompass corrosion depth (Cd), corrosion width (Cw), corrosion length (Cl), void width (Vw), void length (Vl), burial depth (H), traffic load (P), pipe diameter (D), wall thickness (t), bedding modulus (Eb), backfill soil modulus (Es), and groundwater level (hw). The detailed parameter ranges for each variable are provided in Table 3.

    Table 3.  The parameter ranges of each physical variable.
    Physical variable Value Physical variable Value
    Minimum Maximum Minimum Maximum
    Cd (cm) 5 90 P (MPa) 0.5 1.5
    Cw (°) 0 180 D (mm) 300 1200
    Cl (m) 0 10 t (mm) 40 120
    Vw (°) 0 120 Eb (MPa) 6 580
    Vl (m) 0 3 Es (MPa) 5 65
    H (m) 0.5 3 hw/H 0.31 8.85

     | Show Table
    DownLoad: CSV

    These physical variables are randomly generated within the specified ranges, resulting in a dataset of 250 samples. The maximum stress signals are formulated by finite element simulation based on these datasets. Figure 5 illustrates the dependence of the maximum stress on each physical variable. As the maximum stress of concrete sewage pipelines is determined by the 12 physical variables, it is challenging to analyze the relationship between individual variables and stress through simple data fitting. As shown in Figure 5, the relationship between maximum stress and each variable is nonlinear and complex, indicating that this is a multivariate nonlinear regression problem. Existing software and algorithms are unable to efficiently and accurately address this challenge; however, our designed SPNN method is capable of handling it effectively.

    Figure 5.  The dependence of maximum stress signals (y) on corrosion depth (Cd), corrosion width (Cw), corrosion length (Cl), void width (Vw), void length (Vl), burial depth (H), traffic load (P), pipe diameter (D), wall thickness (t), bedding modulus (Eb), backfill soil modulus (Es), and groundwater level over burial depth (hw/H).

    According to the univariate empirical formula [7,9], we define the multivariate regression equation of the maximum stress as

    y= α1Cdβ1+α2Cwβ2+α3Clβ3+α4Vwβ4+α5Vlβ5+α6Hβ6+α7Pβ7+α8Dβ8+α9tβ9+α10Ebβ10+α11Esβ11+α12(hw/H)β12. (3.3)

    The data set is first normalized to eliminate the influence of dimensionality, and then randomly split into training, validation, and test sets, with respective ratios of 64%, 20%, and 16%. It makes almost no difference to divide the data set at the other ratios, for example, 70%, 15%, and 15%, because of the excellent model performance. The regression model is developed using the explainable SPNN. During the training process, a total of 1000 epochs are set, and the Adam optimizer with with a learning rate equal to 0.006 is employed together with a one-cycle scheduler with the maximum learn rate equal to 0.01. Figure 6(a) shows the training loss and valid loss depicted by MSE, and Figure 6(b) presents a comparison between the real and predicted maximum stress values.

    Figure 6.  The multivariate polynomial regression of the maximum stress: (a) the training loss and valid loss depicted by MSE; (b) the plot of real values versus regression values.

    The regression error MSE =3.36×105, and the coefficient of determination R2=0.9998. We save model parameters to two decimal places, and the regression equation is

    y= 0.35Cd0.51+0.24Cw0.32+0.05Cl0.62+0.25Vw0.06+0.28Vl0.38+0.15H0.19+0.92P1.20+0.72D1.16+0.46t1.13+0.37Eb0.240.04Es0.630.04(hw/H)0.43. (3.4)

    It should be noted that the values of coefficients and exponents in Eq (3.4) may exhibit minor fluctuations due to the broad feasible region of the solutions. However, the contributions of each physical variable to the maximum stress derived from Eq (3.4) can be well interpreted. It is evident that the corrosion depth, corrosion width, corrosion length, void width, void length, burial depth, traffic load, pipe diameter, wall thickness, and bedding modulus have positive effects on the maximum stress, while the backfill soil modulus and groundwater level exhibit negative effects. These findings align with existing empirical knowledge.

    Additionally, the SHAP approach [16,17,18] is utilized to analyze the impact of input features on the model's output. The SHAP value for each sample (x) corresponding to feature (fi) is computed using Eq (3.5).

    SHAPfi(x)=fif[|f|×(F|f|)]1×[Pf(x)Pffi(x)], (3.5)

    where fi represents the ith feature (i.e., the ith physical variable), F is the total number of features, and f is a feature subset that includes fi. |f| denotes the number of elements in subset f, and (F|f|) is the number of combinations of F items taken |f| at a time. Pf(x) is the predicted value when all the physical variables in f are selected, and Pffi(x) is the predicted value with all the variables in f except fi. The SHAP values for the samples related to each physical variable are shown in Figure 7.

    Figure 7.  The feature analysis of 12 physical variables on the maximum stress by SHAP values.

    In Figure 7, if the higher feature value corresponds to larger SHAP values, it indicates a positive effect of that feature on the model output. Conversely, if the higher feature value results in smaller SHAP values, it suggests a negative effect. Thus, we conclude that features P,D,t,Cd,Vl,Eb,Cw,Cl,H, and Vw have a positive effect on the maximum stress, while Es and hw/H have a negative effect. These conclusions align with the regression analysis results from Eq (3.4). Additionally, the importance ranking of the input features is P>D>t>Cd>Vl>Eb>Cw>Cl>H>Vw>Es>hw/H.

    According to the regression results, we can further analyze the evolution of the maximum stress of concrete sewage pipelines with respect to the univariate input. By selecting a sample data point and employing the control variable method, the maximum stress signal can be calculated using Eq (3.4), with only one variable being varied. Figure 8 shows the evolution of maximum stress with respect to each physical variable. These univariate evolutionary trends provide theoretical guidance for strategies aimed at repairing and improving the properties of concrete sewage pipelines.

    Figure 8.  The evolution of maximum stress with respect to each physical variable by the control variable method.

    In this manuscript, we propose an explainable sigma-pi neural network to address the multi-variable nonlinear polynomial regression problem. The coefficient and exponent parameters of polynomials are effectively represented by the corresponding weight parameters in the SPNN. To accelerate the regression process, the back-propagation algorithm is employed for parameter optimization. The examples reveal that our designed SPNN can efficiently and accurately solve both integer-order and fractional-order multivariate polynomial regression problems. In practical application, the SPNN provides high-precision fitting of the maximum stress in concrete sewage pipelines under the combined influence of 12 physical parameter variables. Furthermore, feature importance ranking and additional analyses on the relationship between maximum stress and these variables can be conducted based on this explainable machine learning model.

    Building upon the framework for solving multi-variable polynomial regression, future algorithms for addressing more complex nonlinear regression problems could be developed through higher-order neural networks, such as sigma-pi or sigma-pi-sigma networks. This study lays the theoretical foundation for developing generalized tools to solve multi-variable nonlinear regression problems.

    The authors declare that this manuscript is the authors' original work, and they have not used Artificial Intelligence (AI) tools in the creation of this article.

    This work was supported by the China Postdoctoral Science Foundation Funded Project (Grant No. 2022M712902), the Natural Science Foundation of Henan (Grant No. 232300421345), and the Zhongyuan Youth Top Talent Plan (Zhongyuan Youth Postdoctoral Innovative Talents).

    The authors declare no conflict of interest with respect to the research, authorship, and/or publication of this article.



    [1] Thomas WI, Znaniecki F (1976) Chlop polski w Europie I Ameryce. wyd. Polskie Warszawa.
    [2] Glorius B (2021) Social innovation within transnational flows of knowledge: The example of student mobility from Bulgaria. Popul Space Place 27: e2452. https://doi.org/10.1002/psp.2452 doi: 10.1002/psp.2452
    [3] Bakalova M (2021) Education and Migration: The (Non)Return of Better Educated Migrants to Bulgaria. Economic Studies 2021: 166–186.
    [4] Black R, Koser K, Munk K, et al. (2004) Understanding Volun tary Return. Sussex Centre for Migration Research: Home Office, Online Report 50/04. Available from: https://webarchive.nationalarchives.gov.uk/ukgwa/20110220105210/rds.homeoffice.gov.uk/rds/pdfs04/rdsolr5004.pdf.
    [5] Nonchev A, Minchev V, Boshnakov V, et al., (2020) Returning migrants, UNWE Publishing complex.
    [6] Stoilova R, Dimitrova E (2017) Emigration from the Perspective of the School-to-Work Transition in Bulgaria. Czech Sociol Rev 53: 903–933. https://doi.org/10.13060/00380288.2017.53.6.380 doi: 10.13060/00380288.2017.53.6.380
    [7] Cassarino JP (2014) Return Migration and Development: The Significance of Migration Cycles, In: Triandafyllidou A (ed.), Routledge Handbook of Immigration and Refugee Studies. New York: Routledge, 216–222.
    [8] Handbook on Measuring International Migration through Population Censuses (2022) UN Available at: https://unstats.un.org/unsd/demographic-social/Standards-and-Methods/files/Handbooks/international-migration/2022-UNSD-Handbook-Meas-Intern-Migration-E.pdf.
    [9] Aronson E (2011) The social animal, 11 edition, Worth Publishers.
    [10] Misheva M, Getova T, Bakalova M (2020) Returnees' Reintegration "Back at Home" and Sustainability of Return. In Return Migration: European and Bulgarian Perspectives, Conference Proceedings Papers, UNWE.
    [11] Nonchev А, Hristova M (2018) Segmentation of returning migrants. Econ Stud 27: 3–24.
  • This article has been cited by:

    1. László Zoltán Zöldi, Anna Sára Ligeti, Zoltán Csányi, The migratory impact of COVID‐19: The role of time and distances in the migration decisions of Hungarians during the COVID‐19 pandemic, 2024, 1544-8444, 10.1002/psp.2804
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2282) PDF downloads(204) Cited by(1)

Figures and Tables

Figures(3)  /  Tables(8)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog